Search results for "Deep neural networks"
showing 10 items of 18 documents
Automatic Content Analysis of Computer-Supported Collaborative Inquiry-Based Learning Using Deep Networks and Attention Mechanisms
2020
Computer-supported collaborative inquiry-based learning (CSCIL) represents a form of active learning in which students jointly pose questions and investigate them in technology-enhanced settings. Scaffolds can enhance CSCIL processes so that students can complete more challenging problems than they could without scaffolds. Scaffolding CSCIL, however, would optimally adapt to the needs of a specific context, group, and stage of the group's learning process. In CSCIL, the stage of the learning process can be characterized by the inquiry-based learning (IBL) phase (orientation, conceptualization, investigation, conclusion, and discussion). In this presentation, we illustrate the potential of a…
A Neural Network model for the Evaluation of Text Complexity in Italian Language: a Representation Point of View
2018
Abstract The goal of a text simplification system (TS) is to create a new text suited to the characteristics of a reader, with the final goal of making it more understandable.The building of an Automatic Text Simplification System (ATS) cannot be separated from a correct evaluation of the text complexity. In fact the ATS must be capable of understanding if a text should be simplified for the target reader or not. In a previous work we have presented a model capable of classifying Italian sentences based on their complexity level. Our model is a Long Short Term Memory (LSTM) Neural Network capable of learning the features of easy-to-read and complex-to-read sentences autonomously from a anno…
A sentence based system for measuring syntax complexity using a recurrent deep neural network
2018
In this paper we present a deep neural network model capable of inducing the rules that identify the syntax complexity of an Italian sentence. Our system, beyond the ability of choosing if a sentence needs of simplification, gives a score that represent the confidence of the model during the process of decision making which could be representative of the sentence complexity. Experiments have been carried out on one public corpus created specifically for the problem of text-simplification.
Evaluation of Deep Neural Networks for Semantic Segmentation of Prostate in T2W MRI
2020
In this paper, we present an evaluation of four encoder&ndash
Towards a deep-learning-based methodology for supporting satire detection
2021
This paper describes an approach for supporting automatic satire detection through effective deep learning (DL) architecture that has been shown to be useful for addressing sarcasm/irony detection problems. We both trained and tested the system exploiting articles derived from two important satiric blogs, Lercio and IlFattoQuotidiano, and significant Italian newspapers.
Computational Modeling of Human Visual Function using Psychophysics, Deep Neural Networks, and Information Theory
2023
Visual perception is a key to unlocking the secrets of brain functions because most of the information is processed through the early visual system and then transmitted to the high-level cognitive perception brain regions. The brain functions as a self-organizing, bio-dynamic, and chaotic system that receives outside information and then decomposes it into pieces of information that can be processed efficiently and independently. The work connects natural image statistics, psychophysics, deep neural networks, and information theory to perceptual vision systems to explore how vision processes information from the outside world and how the information coupled drives functional connectivity be…
A Novel Systolic Parallel Hardware Architecture for the FPGA Acceleration of Feedforward Neural Networks
2019
New chips for machine learning applications appear, they are tuned for a specific topology, being efficient by using highly parallel designs at the cost of high power or large complex devices. However, the computational demands of deep neural networks require flexible and efficient hardware architectures able to fit different applications, neural network types, number of inputs, outputs, layers, and units in each layer, making the migration from software to hardware easy. This paper describes novel hardware implementing any feedforward neural network (FFNN): multilayer perceptron, autoencoder, and logistic regression. The architecture admits an arbitrary input and output number, units in la…
Multi-class Text Complexity Evaluation via Deep Neural Networks
2019
Automatic Text Complexity Evaluation (ATE) is a natural language processing task which aims to assess texts difficulty taking into account many facets related to complexity. A large number of papers tackle the problem of ATE by means of machine learning algorithms in order to classify texts into complex or simple classes. In this paper, we try to go beyond the methodologies presented so far by introducing a preliminary system based on a deep neural network model whose objective is to classify sentences into more of two classes. Experiments have been carried out on a manually annotated corpus which has been preprocessed in order to make it suitable for the scope of the paper. The results sho…
Highly Performant, Deep Neural Networks with sub-microsecond latency on FPGAs for Trigger Applications
2020
Artificial neural networks are becoming a standard tool for data analysis, but their potential remains yet to be widely used for hardware-level trigger applications. Nowadays, high-end FPGAs, often used in low-level hardware triggers, offer theoretically enough performance to include networks of considerable size. This makes it very promising and rewarding to optimize a neural network implementation for FPGAs in the trigger context. Here an optimized neural network implementation framework is presented, which typically reaches 90 to 100% computational efficiency, requires few extra FPGA resources for data flow and controlling, and allows latencies in the order of 10s to few 100s of nanoseco…
Neural Networks with Multidimensional Cross-Entropy Loss Functions
2019
Deep neural networks have emerged as an effective machine learning tool successfully applied for many tasks, such as misinformation detection, natural language processing, image recognition, machine translation, etc. Neural networks are often applied to binary or multi-class classification problems. In these settings, cross-entropy is used as a loss function for neural network training. In this short note, we propose an extension of the concept of cross-entropy, referred to as multidimensional cross-entropy, and its application as a loss function for classification using neural networks. The presented computational experiments on a benchmark dataset suggest that the proposed approaches may …